-
-
Notifications
You must be signed in to change notification settings - Fork 899
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
fix(config): passing gradient_checkpoint_kwargs #1412
Conversation
In my experience, |
see my comment in #1167 |
In that code block above, it seems that the Then, the Either way, I think one of the defaults should be removed to prevent future confusion. Edit: your linked PR sets it to true, despite the comment saying it's false. |
Thanks for digging into this. Good to go! |
* fix(config): change default use_reentrant to true * Update trainer_builder.py * fix: make sure to pass kwargs to enable checkpoint * chore: lint
* fix(config): change default use_reentrant to true * Update trainer_builder.py * fix: make sure to pass kwargs to enable checkpoint * chore: lint
According to huggingface/transformers#28339 , setting it to
False
increases VRAM. My quick testing shows ~1GB increase at lowest settings.Furthermore, the default in transformers and torch is going to be
True
huggingface/transformers#29638 (comment)Finally, removing the default to False in trainer_building to clean old configs. I see that this kwarg is now set in https://github.com/OpenAccess-AI-Collective/axolotl/blob/a914cb37dc455a3fd0368e3a0898867f25b3a6c9/src/axolotl/utils/config/__init__.py#L170-L176
Description
Motivation and Context
How has this been tested?
Screenshots (if appropriate)
Types of changes
Social Handles (Optional)